Discrete Recurrent Neural Networks as Pushdown Automata

نویسندگان

  • Zheng Zeng
  • Rodney M. Goodman
  • Padhraic Smyth
چکیده

in this paper we describe a new discrete rccurrcnt neural network model with discrete external stacks for learning context-free grammars (or pushdown automata). Conventional analog recurrent networks tend to have stability problems when presented with input sirings which are longer than those used for training: the network’s internal states become merged and the string can not be correctly parsed. However, the discrete recurrent structure forms a stable representation during learning by using isolated discrete points as its internal representation of states for the automata. Hence, once successfully trained, the network is perfectly stable on input strings of arbitrary length. For training such discrete networks a novel “pseudo-gradient” learning rule is used, Experimental results demonstrate the ability of the discrete network to learn context-free grammars in a stable manner. ‘1’hc discrete network model results in the advantages of a stable network, a clear understanding of the operation of the stack, and a structure which is easily irnplcmcntable in hardware.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Discrete recurrent neural networks for grammatical inference

Describes a novel neural architecture for learning deterministic context-free grammars, or equivalently, deterministic pushdown automata. The unique feature of the proposed network is that it forms stable state representations during learning-previous work has shown that conventional analog recurrent networks can be inherently unstable in that they cannot retain their state memory for long inpu...

متن کامل

The Neural Network Pushdown Automaton: Architecture, Dynamics and Training

Recurrent neural networks are dynamical network structures which have the capabilities of processing and generating temporal information. To our knowledge the earliest neural network model that processed temporal information was that of MeCulloch and Pitts [McCulloch43]. Kleene [Kleene56] extended this work to show the equivalence of finite automata and McCulloch and Pitts' representation of ne...

متن کامل

The expressive power of analog recurrent neural networks on infinite input streams

We consider analog recurrent neural networks working on infinite input streams, provide a complete topological characterization of their expressive power, and compare it to the expressive power of classical infinite word reading abstract machines. More precisely, we consider analog recurrent neural networks as language recognizers over the Cantor space, and prove that the classes of ω-languages...

متن کامل

The Neural Network Pushdown Automaton: Model, Stack and Learning Simulations

In order for neural networks to learn complex languages or grammars, they must have sufficient computational power or resources to recognize or generate such languages. Though many approaches have been discussed, one obvious approach to enhancing the processing power of a recurrent neural network is to couple it with an external stack memory in effect creating a neural network pushdown automata...

متن کامل

The Equivalence Problem of Deterministic Multitape Finite Automata: A New Proof of Solvability Using a Multidimensional Tape

Complexity in Convex Languages p. 1 Three Learnable Models for the Description of Language p. 16 Arbology: Trees and Pushdown Automata p. 32 Analysis of Communicating Automata p. 50 Regular Papers Complexity of the Satisfiability Problem for a Class of Propositional Schemata p. 58 A Simple n-Dimensional Intrinsically Universal Quantum Cellular Automaton p. 70 A Fast Longest Common Subsequence A...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1997